navigation module
c6e954799a0218f6d341ad5cbfb58999-Paper-Conference.pdf
Invideo recognition, weneedtosample multiple frames torepresent eachvideo which makesthe computational cost scale proportionally to the number of sampled frames. In most cases, a small proportion of all the frames is sampled for each input, which only contains limited information of the original video.
Towards an Autonomous Compost Turner: Current State of Research
Cichocki, Max, Reitbauer, Eva, Theurl, Fabian, Schmied, Christoph
This preprint presents the current status of research into the development and application of an autonomous, self-driving compost turner. The aim is to overcome challenges in the composting industry, such as adverse working conditions, by automating the composting process. The preprint provides a comprehensive overview of the overall concept of the self-driving compost turner, including the hardware architecture with sensors, navigation module and control module. In addition, the methodical development of the architecture of concepts, models and their subsequent software integration in ROS using model-based systems engineering is described. The validation and verification of the overall system is carried out in an industrial environment using three scenarios. The capabilities of the compost turner are demonstrated by autonomously following predefined trajectories in the composting plant and performing the required composting tasks. The results show that the autonomous compost turner is capable of performing the required activities. In addition, the compost turner has intelligent processing capabilities for compost data as well as its transmission, visualization and storage in a cloud server. It is important to note that this work is a preprint that represents the current state of research. The authors aim to publish the full paper in a peer-reviewed journal in the near future.
- Europe > Austria > Styria > Graz (0.07)
- North America > United States > Michigan (0.04)
- North America > United States > California > San Diego County > San Diego (0.04)
- (2 more...)
- Workflow (1.00)
- Research Report > New Finding (0.34)
- Information Technology (1.00)
- Energy (1.00)
- Food & Agriculture > Agriculture (0.93)
A Map-Free LiDAR-Based System for Autonomous Navigation in Vineyards
Bertoglio, Riccardo, Carini, Veronica, Arrigoni, Stefano, Matteucci, Matteo
Agricultural robots have the potential to increase production yields and reduce costs by performing repetitive and time-consuming tasks. However, for robots to be effective, they must be able to navigate autonomously in fields or orchards without human intervention. In this paper, we introduce a navigation system that utilizes LiDAR and wheel encoder sensors for in-row, turn, and end-row navigation in row structured agricultural environments, such as vineyards. Our approach exploits the simple and precise geometrical structure of plants organized in parallel rows. We tested our system in both simulated and real environments, and the results demonstrate the effectiveness of our approach in achieving accurate and robust navigation. Our navigation system achieves mean displacement errors from the center line of 0.049 m and 0.372 m for in-row navigation in the simulated and real environments, respectively. In addition, we developed an end-row points detection that allows end-row navigation in vineyards, a task often ignored by most works.
ANYmal Parkour: Learning Agile Navigation for Quadrupedal Robots
Hoeller, David, Rudin, Nikita, Sako, Dhionis, Hutter, Marco
Performing agile navigation with four-legged robots is a challenging task due to the highly dynamic motions, contacts with various parts of the robot, and the limited field of view of the perception sensors. In this paper, we propose a fully-learned approach to train such robots and conquer scenarios that are reminiscent of parkour challenges. The method involves training advanced locomotion skills for several types of obstacles, such as walking, jumping, climbing, and crouching, and then using a high-level policy to select and control those skills across the terrain. Thanks to our hierarchical formulation, the navigation policy is aware of the capabilities of each skill, and it will adapt its behavior depending on the scenario at hand. Additionally, a perception module is trained to reconstruct obstacles from highly occluded and noisy sensory data and endows the pipeline with scene understanding. Compared to previous attempts, our method can plan a path for challenging scenarios without expert demonstration, offline computation, a priori knowledge of the environment, or taking contacts explicitly into account. While these modules are trained from simulated data only, our real-world experiments demonstrate successful transfer on hardware, where the robot navigates and crosses consecutive challenging obstacles with speeds of up to two meters per second. The supplementary video can be found on the project website: https://sites.google.com/leggedrobotics.com/agile-navigation
- Europe > Switzerland > Zürich > Zürich (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Europe > Germany > Baden-Württemberg > Freiburg (0.04)
- Asia > Middle East > Republic of Türkiye > Karaman Province > Karaman (0.04)
Counterfactual Fairness Filter for Fair-Delay Multi-Robot Navigation
Asano, Hikaru, Yonetani, Ryo, Nishimura, Mai, Kozuno, Tadashi
Multi-robot navigation is the task of finding trajectories for a team of robotic agents to reach their destinations as quickly as possible without collisions. In this work, we introduce a new problem: fair-delay multi-robot navigation, which aims not only to enable such efficient, safe travels but also to equalize the travel delays among agents in terms of actual trajectories as compared to the best possible trajectories. The learning of a navigation policy to achieve this objective requires resolving a nontrivial credit assignment problem with robotic agents having continuous action spaces. Hence, we developed a new algorithm called Navigation with Counterfactual Fairness Filter (NCF2). With NCF2, each agent performs counterfactual inference on whether it can advance toward its goal or should stay still to let other agents go. Doing so allows us to effectively address the aforementioned credit assignment problem and improve fairness regarding travel delays while maintaining high efficiency and safety. Our extensive experimental results in several challenging multi-robot navigation environments demonstrate the greater effectiveness of NCF2 as compared to state-of-the-art fairness-aware multi-agent reinforcement learning methods. Our demo videos and code are available on the project webpage: https://omron-sinicx.github.io/ncf2/
- Information Technology > Artificial Intelligence > Robots (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Undirected Networks > Markov Models (0.47)
Look More but Care Less in Video Recognition
Zhang, Yitian, Bai, Yue, Wang, Huan, Xu, Yi, Fu, Yun
Existing action recognition methods typically sample a few frames to represent each video to avoid the enormous computation, which often limits the recognition performance. To tackle this problem, we propose Ample and Focal Network (AFNet), which is composed of two branches to utilize more frames but with less computation. Specifically, the Ample Branch takes all input frames to obtain abundant information with condensed computation and provides the guidance for Focal Branch by the proposed Navigation Module; the Focal Branch squeezes the temporal size to only focus on the salient frames at each convolution block; in the end, the results of two branches are adaptively fused to prevent the loss of information. With this design, we can introduce more frames to the network but cost less computation. Besides, we demonstrate AFNet can utilize fewer frames while achieving higher accuracy as the dynamic selection in intermediate features enforces implicit temporal modeling. Further, we show that our method can be extended to reduce spatial redundancy with even less cost. Extensive experiments on five datasets demonstrate the effectiveness and efficiency of our method.
Adversarial Attack Against Image-Based Localization Neural Networks
Brand, Meir, Naeh, Itay, Teitelman, Daniel
In this paper, we present a proof of concept for adversarially attacking the image-based localization module of an autonomous vehicle. This attack aims to cause the vehicle to perform a wrong navigational decisions and prevent it from reaching a desired predefined destination in a simulated urban environment. A database of rendered images allowed us to train a deep neural network that performs a localization task and implement, develop and assess the adversarial pattern. Our tests show that using this adversarial attack we can prevent the vehicle from turning at a given intersection. This is done by manipulating the vehicle's navigational module to falsely estimate its current position and thus fail to initialize the turning procedure until the vehicle misses the last opportunity to perform a safe turn in a given intersection.
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- Asia > Middle East > Israel (0.04)
- Asia > India (0.04)
- Government > Military (0.72)
- Transportation > Ground > Road (0.69)
- Information Technology > Security & Privacy (0.62)
Learning to Visually Navigate in Photorealistic Environments Without any Supervision
Mezghani, Lina, Sukhbaatar, Sainbayar, Szlam, Arthur, Joulin, Armand, Bojanowski, Piotr
Learning to navigate in a realistic setting where an agent must rely solely on visual inputs is a challenging task, in part because the lack of position information makes it difficult to provide supervision during training. In this paper, we introduce a novel approach for learning to navigate from image inputs without external supervision or reward. Our approach consists of three stages: learning a good representation of first-person views, then learning to explore using memory, and finally learning to navigate by setting its own goals. The model is trained with intrinsic rewards only so that it can be applied to any environment with image observations. We show the benefits of our approach by training an agent to navigate challenging photo-realistic environments from the Gibson dataset with RGB inputs only.
REAPER: A Reflexive Architecture for Perceptive Agents
Maxwell, Bruce A., Meeden, Lisa A., Addo, Nii Saka, Dickson, Paul, Fairfield, Nathaniel, Johnson, Nikolas, Jones, Edward G., Kim, Suor, Malla, Pukar, Murphy, Matthew, Rutter, Brandon, Silk, Eli
This article describes the winning entries in the 2000 Association for the Advancement of Artificial Intelligence Mobile Robot Competition. The robots, developed by Swarthmore College, all used a modular hybrid architecture designed to enable reflexive responses to perceptual input. Within this architecture, the robots integrated visual sensing, speech synthesis and recognition, the display of an animated face, navigation, and interrobot communication. In the Hors d'Oeuvres, Anyone? event, a team of robots entertained the crowd while they interactively served cookies; and in the Urban Search-and-Rescue event, a single robot autonomously explored a section of the test area, identified interesting features, built an annotated map, and exited the test area within the allotted time.
- North America > United States > Massachusetts > Middlesex County > Reading (0.04)
- North America > United States > Virginia (0.04)
- North America > United States > Massachusetts > Norfolk County > Wellesley (0.04)
- (6 more...)